@InProceedings{LimaKuxShig:2018:AcAuNa,
author = "Lima, Sidney Andrade de and Kux, Hermann Johann Heinrich and
Shiguemori, Elcio Hideiti",
affiliation = "{Instituto de Estudos Avan{\c{c}}ados (IEAv)} and {Instituto
Nacional de Pesquisas Espaciais (INPE)} and {Instituto de Estudos
Avan{\c{c}}ados (IEAv)}",
title = "Accuracy of Autonomy Navigation of Unmanned Aircraft Systems
through Imagery",
booktitle = "Proceedings...",
year = "2018",
pages = "848--852",
organization = "20th International Conference on Unmanned Aircraft Systems -
ICUAS-2018.",
publisher = "World Academy of Science, Engineering and Technology - WASET",
keywords = "Autonomy, navigation, security, photogrammetry, UAS.",
abstract = "The Unmanned Aircraft Systems (UAS) usually navigate through the
Global Navigation Satellite System (GNSS) associated with an
Inertial Navigation System (INS). However, GNSS can have its
accuracy degraded at any time or even turn off the signal of GNSS.
In addition, there is the possibility of malicious interferences,
known as jamming. Therefore, the image navigation system can solve
the autonomy problem, because if the GNSS is disabled or degraded,
the image navigation system would continue to provide coordinate
information for the INS, allowing the autonomy of the system. This
work aims to evaluate the accuracy of the positioning though
photogrammetry concepts. The methodology uses orthophotos and
Digital Surface Models (DSM) as a reference to represent the
object space and photograph obtained during the flight to
represent the image space. For the calculation of the coordinates
of the perspective center and camera attitudes, it is necessary to
know the coordinates of homologous points in the object space
(orthophoto coordinates and DSM altitude) and image space (column
and line of the photograph). So if it is possible to automatically
identify in real time the homologous points the coordinates and
attitudes can be calculated whit their respective accuracies. With
the methodology applied in this work, it is possible to verify
maximum errors in the order of 0.5 m in the positioning and 0.6º
in the attitude of the camera, so the navigation through the image
can reach values equal to or higher than the GNSS receivers
without differential correction. Therefore, navigating through the
image is a good alternative to enable autonomous navigation.",
conference-location = "Amsterdam Amsterdam",
conference-year = "2018",
isbn = "1307-6892",
label = "lattes: 5178191377737639 1 LimaKuxShig:2018:AcAuNa",
language = "en",
targetfile = "lima_accuracy.pdf",
url = "https://waset.org/Publications/accuracy-of-autonomy-navigation-of-unmanned-aircraft-systems-through-imagery/10008886",
volume = "12",
urlaccessdate = "27 abr. 2024"
}